|
In mathematics, approximation theory is concerned with how functions can best be approximated with simpler functions, and with quantitatively characterizing the errors introduced thereby. Note that what is meant by ''best'' and ''simpler'' will depend on the application. A closely related topic is the approximation of functions by generalized Fourier series, that is, approximations based upon summation of a series of terms based upon orthogonal polynomials. One problem of particular interest is that of approximating a function in a computer mathematical library, using operations that can be performed on the computer or calculator (e.g. addition and multiplication), such that the result is as close to the actual function as possible. This is typically done with polynomial or rational (ratio of polynomials) approximations. The objective is to make the approximation as close as possible to the actual function, typically with an accuracy close to that of the underlying computer's floating point arithmetic. This is accomplished by using a polynomial of high degree, and/or narrowing the domain over which the polynomial has to approximate the function. Narrowing the domain can often be done through the use of various addition or scaling formulas for the function being approximated. Modern mathematical libraries often reduce the domain into many tiny segments and use a low-degree polynomial for each segment. ==Optimal polynomials== Once the domain (typically an interval) and degree of the polynomial are chosen, the polynomial itself is chosen in such a way as to minimize the worst-case error. That is, the goal is to minimize the maximum value of , where ''P''(''x'') is the approximating polynomial, ''f''(''x'') is the actual function, and ''x'' varies over the chosen interval. For well-behaved functions, there exists an ''N''th-degree polynomial that will lead to an error curve that oscillates back and forth between and a total of ''N''+2 times, giving a worst-case error of . It is seen that an ''N''th-degree polynomial can interpolate ''N''+1 points in a curve. Such a polynomial is always optimal. It is possible to make contrived functions ''f''(''x'') for which no such polynomial exists, but these occur rarely in practice. For example, the graphs shown to the right show the error in approximating log(x) and exp(x) for ''N'' = 4. The red curves, for the optimal polynomial, are level, that is, they oscillate between and exactly. Note that, in each case, the number of extrema is ''N''+2, that is, 6. Two of the extrema are at the end points of the interval, at the left and right edges of the graphs. To prove this is true in general, suppose ''P'' is a polynomial of degree ''N'' having the property described, that is, it gives rise to an error function that has ''N'' + 2 extrema, of alternating signs and equal magnitudes. The red graph to the right shows what this error function might look like for ''N'' = 4. Suppose ''Q''(''x'') (whose error function is shown in blue to the right) is another ''N''-degree polynomial that is a better approximation to ''f'' than ''P''. In particular, ''Q'' is closer to ''f'' than ''P'' for each value ''xi'' where an extreme of ''P''−''f'' occurs, so : When a maximum of ''P''−''f'' occurs at ''xi'', then : And when a minimum of ''P''−''f'' occurs at ''xi'', then : So, as can be seen in the graph, () − () must alternate in sign for the ''N'' + 2 values of ''xi''. But () − () reduces to ''P''(''x'') − ''Q''(''x'') which is a polynomial of degree ''N''. This function changes sign at least ''N''+1 times so, by the Intermediate value theorem, it has ''N''+1 zeroes, which is impossible for a polynomial of degree ''N''. 抄文引用元・出典: フリー百科事典『 ウィキペディア(Wikipedia)』 ■ウィキペディアで「approximation theory」の詳細全文を読む スポンサード リンク
|